Back To Index Previous Article Next Article Full Text


Statistica Sinica 1(1991), 127-136


A BIAS BOUND FOR LEAST SQUARES

LINEAR REGRESSION


Naihua Duan and Ker-Chau Li


RAND Corporation and University of California


Abstract: Consider a general linear model y=g(α+βx)+ε , where the link function g is arbitrary and unknown. The maximal component of (λ, β) that can be identified is the direction of β, which measures the substitutibility of the components of x. If ζ(βx)=E(x|βx) is linear in βx, the least squares linear regression of y on x gives a consistent estimate for the direction of β, despite possible nonlinearity in the link function (Brillinger (1977, 1982)). If ζ(βx) is nonlinear, the linear regreesion might be inconsistent for the direction of β. We establish a bound for the asymptotic bias, which is determined from the nonlinearity in ζ(βx) , and the multiple correlation coefficient R2 for the least squares linear regression of y on x. According to the bias bound, the linear regression is nearly consistent for the direction of β, despite possible nonlinearity in the link function, provided that the nonlinearity in ζ(βx) is small compared to R2. Our measure of nonlinearity in ζ(βx) is analogous to the maximal curvature studied by Cox and Small (1978). The bias bound is tight; we give the construction for the least favorable models which achieve the bias bound. The theory is applied to a special case for an illustration.



Key words and phrases: Lack of fit, link function, maximal curvature, nonlinearity, projection index, projection pursuit.



Back To Index Previous Article Next Article Full Text